in alloc_vm_area().
Signed-off-by: Keir Fraser <keir@xensource.com>
}
EXPORT_SYMBOL_GPL(free_vm_area);
-void lock_vm_area(struct vm_struct *area)
-{
- // nothing
-}
-EXPORT_SYMBOL_GPL(lock_vm_area);
-
-void unlock_vm_area(struct vm_struct *area)
-{
- // nothing
-}
-EXPORT_SYMBOL_GPL(unlock_vm_area);
-
/*
* Local variables:
* c-file-style: "linux"
static int map_frontend_page(blkif_t *blkif, unsigned long shared_page)
{
struct gnttab_map_grant_ref op;
- int ret;
gnttab_set_map_op(&op, (unsigned long)blkif->blk_ring_area->addr,
GNTMAP_host_map, shared_page, blkif->domid);
- lock_vm_area(blkif->blk_ring_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1);
- unlock_vm_area(blkif->blk_ring_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
+ BUG();
if (op.status) {
DPRINTK(" Grant table operation failure !\n");
static void unmap_frontend_page(blkif_t *blkif)
{
struct gnttab_unmap_grant_ref op;
- int ret;
gnttab_set_unmap_op(&op, (unsigned long)blkif->blk_ring_area->addr,
GNTMAP_host_map, blkif->shmem_handle);
- lock_vm_area(blkif->blk_ring_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1);
- unlock_vm_area(blkif->blk_ring_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
+ BUG();
}
int blkif_map(blkif_t *blkif, unsigned long shared_page, unsigned int evtchn)
static int map_frontend_page(blkif_t *blkif, unsigned long shared_page)
{
struct gnttab_map_grant_ref op;
- int ret;
gnttab_set_map_op(&op, (unsigned long)blkif->blk_ring_area->addr,
GNTMAP_host_map, shared_page, blkif->domid);
- lock_vm_area(blkif->blk_ring_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1);
- unlock_vm_area(blkif->blk_ring_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
+ BUG();
if (op.status) {
DPRINTK(" Grant table operation failure !\n");
static void unmap_frontend_page(blkif_t *blkif)
{
struct gnttab_unmap_grant_ref op;
- int ret;
gnttab_set_unmap_op(&op, (unsigned long)blkif->blk_ring_area->addr,
GNTMAP_host_map, blkif->shmem_handle);
- lock_vm_area(blkif->blk_ring_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1);
- unlock_vm_area(blkif->blk_ring_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
+ BUG();
}
int tap_blkif_map(blkif_t *blkif, unsigned long shared_page,
netif_t *netif, grant_ref_t tx_ring_ref, grant_ref_t rx_ring_ref)
{
struct gnttab_map_grant_ref op;
- int ret;
gnttab_set_map_op(&op, (unsigned long)netif->tx_comms_area->addr,
GNTMAP_host_map, tx_ring_ref, netif->domid);
- lock_vm_area(netif->tx_comms_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1);
- unlock_vm_area(netif->tx_comms_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
+ BUG();
if (op.status) {
DPRINTK(" Gnttab failure mapping tx_ring_ref!\n");
gnttab_set_map_op(&op, (unsigned long)netif->rx_comms_area->addr,
GNTMAP_host_map, rx_ring_ref, netif->domid);
- lock_vm_area(netif->rx_comms_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1);
- unlock_vm_area(netif->rx_comms_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
+ BUG();
if (op.status) {
DPRINTK(" Gnttab failure mapping rx_ring_ref!\n");
static void unmap_frontend_pages(netif_t *netif)
{
struct gnttab_unmap_grant_ref op;
- int ret;
gnttab_set_unmap_op(&op, (unsigned long)netif->tx_comms_area->addr,
GNTMAP_host_map, netif->tx_shmem_handle);
- lock_vm_area(netif->tx_comms_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1);
- unlock_vm_area(netif->tx_comms_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
+ BUG();
gnttab_set_unmap_op(&op, (unsigned long)netif->rx_comms_area->addr,
GNTMAP_host_map, netif->rx_shmem_handle);
- lock_vm_area(netif->rx_comms_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1);
- unlock_vm_area(netif->rx_comms_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
+ BUG();
}
int netif_map(netif_t *netif, unsigned long tx_ring_ref,
static int map_frontend_page(tpmif_t *tpmif, unsigned long shared_page)
{
- int ret;
struct gnttab_map_grant_ref op;
gnttab_set_map_op(&op, (unsigned long)tpmif->tx_area->addr,
GNTMAP_host_map, shared_page, tpmif->domid);
- lock_vm_area(tpmif->tx_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1);
- unlock_vm_area(tpmif->tx_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
+ BUG();
if (op.status) {
DPRINTK(" Grant table operation failure !\n");
static void unmap_frontend_page(tpmif_t *tpmif)
{
struct gnttab_unmap_grant_ref op;
- int ret;
gnttab_set_unmap_op(&op, (unsigned long)tpmif->tx_area->addr,
GNTMAP_host_map, tpmif->shmem_handle);
- lock_vm_area(tpmif->tx_area);
- ret = HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1);
- unlock_vm_area(tpmif->tx_area);
- BUG_ON(ret);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
+ BUG();
}
int tpmif_map(tpmif_t *tpmif, unsigned long shared_page, unsigned int evtchn)
return NULL;
}
+ /* Map page directories into every address space. */
+ vmalloc_sync_all();
+
return area;
}
EXPORT_SYMBOL_GPL(alloc_vm_area);
kfree(area);
}
EXPORT_SYMBOL_GPL(free_vm_area);
-
-void lock_vm_area(struct vm_struct *area)
-{
- unsigned long i;
- char c;
-
- /*
- * Prevent context switch to a lazy mm that doesn't have this area
- * mapped into its page tables.
- */
- preempt_disable();
-
- /*
- * Ensure that the page tables are mapped into the current mm. The
- * page-fault path will copy the page directory pointers from init_mm.
- */
- for (i = 0; i < area->size; i += PAGE_SIZE)
- (void)__get_user(c, (char __user *)area->addr + i);
-}
-EXPORT_SYMBOL_GPL(lock_vm_area);
-
-void unlock_vm_area(struct vm_struct *area)
-{
- preempt_enable();
-}
-EXPORT_SYMBOL_GPL(unlock_vm_area);
gnttab_set_map_op(&op, (unsigned long)area->addr, GNTMAP_host_map,
gnt_ref, dev->otherend_id);
- lock_vm_area(area);
- BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1));
- unlock_vm_area(area);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
+ BUG();
if (op.status != GNTST_okay) {
free_vm_area(area);
gnttab_set_map_op(&op, (unsigned long)vaddr, GNTMAP_host_map,
gnt_ref, dev->otherend_id);
- BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1));
+ if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
+ BUG();
if (op.status != GNTST_okay) {
xenbus_dev_fatal(dev, op.status,
gnttab_set_unmap_op(&op, (unsigned long)area->addr, GNTMAP_host_map,
(grant_handle_t)area->phys_addr);
- lock_vm_area(area);
- BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1));
- unlock_vm_area(area);
+ if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
+ BUG();
if (op.status == GNTST_okay)
free_vm_area(area);
gnttab_set_unmap_op(&op, (unsigned long)vaddr, GNTMAP_host_map,
handle);
- BUG_ON(HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1));
+ if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
+ BUG();
if (op.status != GNTST_okay)
xenbus_dev_error(dev, op.status,
extern struct vm_struct *alloc_vm_area(unsigned long size);
extern void free_vm_area(struct vm_struct *area);
-/* Lock an area so that PTEs are accessible in the current address space. */
-extern void lock_vm_area(struct vm_struct *area);
-extern void unlock_vm_area(struct vm_struct *area);
-
#endif /* __ASM_XEN_DRIVER_UTIL_H__ */